T\"urkiye is located on a fault line; earthquakes often occur on a large and small scale. There is a need for effective solutions for gathering current information during disasters. We can use social media to get insight into public opinion. This insight can be used in public relations and disaster management. In this study, Twitter posts on Izmir Earthquake that took place on October 2020 are analyzed. We question if this analysis can be used to make social inferences on time. Data mining and natural language processing (NLP) methods are used for this analysis. NLP is used for sentiment analysis and topic modelling. The latent Dirichlet Allocation (LDA) algorithm is used for topic modelling. We used the Bidirectional Encoder Representations from Transformers (BERT) model working with Transformers architecture for sentiment analysis. It is shown that the users shared their goodwill wishes and aimed to contribute to the initiated aid activities after the earthquake. The users desired to make their voices heard by competent institutions and organizations. The proposed methods work effectively. Future studies are also discussed.
translated by 谷歌翻译
In this paper, we introduce a novel optimization algorithm for machine learning model training called Normalized Stochastic Gradient Descent (NSGD) inspired by Normalized Least Mean Squares (NLMS) from adaptive filtering. When we train a high-complexity model on a large dataset, the learning rate is significantly important as a poor choice of optimizer parameters can lead to divergence. The algorithm updates the new set of network weights using the stochastic gradient but with $\ell_1$ and $\ell_2$-based normalizations on the learning rate parameter similar to the NLMS algorithm. Our main difference from the existing normalization methods is that we do not include the error term in the normalization process. We normalize the update term using the input vector to the neuron. Our experiments present that the model can be trained to a better accuracy level on different initial settings using our optimization algorithm. In this paper, we demonstrate the efficiency of our training algorithm using ResNet-20 and a toy neural network on different benchmark datasets with different initializations. The NSGD improves the accuracy of the ResNet-20 from 91.96\% to 92.20\% on the CIFAR-10 dataset.
translated by 谷歌翻译
Recent 3D-aware GANs rely on volumetric rendering techniques to disentangle the pose and appearance of objects, de facto generating entire 3D volumes rather than single-view 2D images from a latent code. Complex image editing tasks can be performed in standard 2D-based GANs (e.g., StyleGAN models) as manipulation of latent dimensions. However, to the best of our knowledge, similar properties have only been partially explored for 3D-aware GAN models. This work aims to fill this gap by showing the limitations of existing methods and proposing LatentSwap3D, a model-agnostic approach designed to enable attribute editing in the latent space of pre-trained 3D-aware GANs. We first identify the most relevant dimensions in the latent space of the model controlling the targeted attribute by relying on the feature importance ranking of a random forest classifier. Then, to apply the transformation, we swap the top-K most relevant latent dimensions of the image being edited with an image exhibiting the desired attribute. Despite its simplicity, LatentSwap3D provides remarkable semantic edits in a disentangled manner and outperforms alternative approaches both qualitatively and quantitatively. We demonstrate our semantic edit approach on various 3D-aware generative models such as pi-GAN, GIRAFFE, StyleSDF, MVCGAN, EG3D and VolumeGAN, and on diverse datasets, such as FFHQ, AFHQ, Cats, MetFaces, and CompCars. The project page can be found: \url{https://enisimsar.github.io/latentswap3d/}.
translated by 谷歌翻译
为了在医学成像研究中保持标准,图像应具有必要的图像质量,以进行潜在的诊断使用。尽管基于CNN的方法用于评估图像质量,但仍可以从准确性方面提高其性能。在这项工作中,我们通过使用SWIN Transformer来解决此问题,这改善了导致医疗图像质量降解的质量质量差分类性能。我们在胸部X射线(Object-CXR)和心脏MRI上的左心室流出路分类问题(LVOT)上测试了胸部X射线(Object-CXR)和左心室流出路分类问题的方法。虽然我们在Object-CXR和LVOT数据集中获得了87.1%和95.48%的分类精度,但我们的实验结果表明,SWIN Transformer的使用可以改善对象CXR分类性能,同时获得LVOT数据集的可比性能。据我们所知,我们的研究是医学图像质量评估的第一个Vision Transformer应用程序。
translated by 谷歌翻译
卷积一直是现代深层神经网络的核心运作。众所周知,可以在傅立叶变换域中实现卷积。在本文中,我们建议使用二进制块WALSH-HATAMARD变换(WHT)而不是傅里叶变换。我们使用基于WHT的二进制层来替换深度神经网络中的一些常规卷积层。我们本文利用了一维(1-D)和二维(2-D)二进制WHT。在两个1-D和2-D层中,我们计算输入特征图的二进制WHT,并使用非线性去噪该WHT域系数,该非线性通过将软阈值与TanH函数组合而获得的非线性。在去噪后,我们计算反相WHT。我们使用1d-wht来取代$ 1 \ times 1 $卷积层,2d-wht层可以取代3 $ \ times $ 3卷积层和挤压和激发层。具有可培训重量的2D-WHT层也可以在全局平均池(间隙)层之前插入以辅助致密层。通过这种方式,我们可以显着降低可训练参数的衡量参数的数量。在本文中,我们将WHT层实施到MobileNet-V2,MobileNet-V3大,并重新阅读,以显着降低参数的数量,以可忽略不计的精度损失。此外,根据我们的速度测试,2D-FWWHT层的运行大约是常规3美元3美元3美元的速度大约为19.51次较少的RAM使用率在NVIDIA Jetson Nano实验中的使用率。
translated by 谷歌翻译
随着深度学习的出现,估计来自单个RGB图像的深度最近受到了很多关注,能够赋予许多不同的应用,从用于计算电影的机器人的路径规划范围。尽管如此,虽然深度地图完全可靠,但对象不连续的估计仍然远非令人满意。这可以有助于卷积运营商自然地聚集在对象不连续性的特征的事实中,导致平滑的过渡而不是明确的边界。因此,为了规避这个问题,我们提出了一种新颖的卷积运营商,明确地定制,以避免不同对象部件的特征聚合。特别地,我们的方法基于借助于超像素估计每个部分深度值。所提出的卷积运算符,我们将“实例卷积”,然后仅在估计的超像素的基础上单独考虑每个对象部分。我们对NYUV2以及IBIMS数据集的评估清楚地展示了在估计遮挡边界周围估算深度的经典卷积上的实例卷积的优越性,同时在其他地方产生了可比结果。代码将在接受时公开提供。
translated by 谷歌翻译